Spectral proximal method for solving large scale sparse optimization
نویسندگان
چکیده
In this paper, we propose to use spectral proximal method solve sparse optimization problems. Sparse refers an problem involving the ι 0 -norm in objective or constraints. The previous research showed that gradient is outperformed other standard unconstrained methods. This due replaced full rank matrix by a diagonal and memory decreased from Ο(n 2 ) Ο(n). Since term nonconvex non-smooth, it cannot be solved algorithm. We will with underdetermined system as its constraint considered. Using Lagrange method, transformed into problem. A new called proposed, which combination of method. then applied programming code written Python compare efficiency proposed some existing benchmarks comparison are based on number iterations, functions call computational time. Theoretically, requires less storage
منابع مشابه
Solving Large-Scale Sparse Semidefinite Programs for Combinatorial Optimization
We present a dual-scaling interior-point algorithm and show how it exploits the structure and sparsity of some large scale problems. We solve the positive semideenite relaxation of combinatorial and quadratic optimization problems subject to boolean constraints. We report the rst computational results of interior-point algorithms for approximating the maximum cut semideenite programs with dimen...
متن کاملA Multilevel Proximal Gradient Algorithm for Large Scale Optimization
Composite optimization models consist of the minimization of the sum of a smooth (not necessarily convex) function and a non-smooth convex function. Such models arise in many applications where, in addition to the composite nature of the objective function, a hierarchy of models is readily available. It is common to take advantage of this hierarchy of models by first solving a low fidelity mode...
متن کاملA Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization
Composite convex optimization models consist of the minimization of the sum of a smooth convex function and a non-smooth convex function. Such models arise in many applications where, in addition to the composite nature of the objective function, a hierarchy of models is readily available. It is common to take advantage of this hierarchy of models by first solving a low fidelity model and then ...
متن کاملIncremental proximal methods for large scale convex optimization
Abstract We consider the minimization of a sum Pm i=1 fi(x) consisting of a large number of convex component functions fi. For this problem, incremental methods consisting of gradient or subgradient iterations applied to single components have proved very effective. We propose new incremental methods, consisting of proximal iterations applied to single components, as well as combinations of gra...
متن کاملSpectral residual method without gradient information for solving large-scale nonlinear systems of equations
A fully derivative-free spectral residual method for solving largescale nonlinear systems of equations is presented. It uses in a systematic way the residual vector as a search direction, a spectral steplength that produces a nonmonotone process and a globalization strategy that allows for this nonmonotone behavior. The global convergence analysis of the combined scheme is presented. An extensi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ITM web of conferences
سال: 2021
ISSN: ['2271-2097', '2431-7578']
DOI: https://doi.org/10.1051/itmconf/20213604007